Towards Efficient Data Movement at Extreme Scale

نویسندگان

  • Yong Chen
  • Yu Zhuang
  • William L. Hase
  • Alan Sill
  • Mark A. Sheridan
  • Yin Lu
  • JiaLin Liu
  • Prathamesh Amritkar
  • Chao Chen
  • JunJie Chen
  • Dong Dai
  • Wei Xie
چکیده

High Performance Computing (HPC) systems are equipped with a massive number of processors, high-bandwidth networks, large capacity fast storage, and specialized parallel software to provide answers to challenging scientific and engineering questions. Over the past decades, data generation capabilities in these domains have grown rapidly due to the emergence of large-scale instruments such as telescopes, colliders, as well as the proliferation of sensors and high-throughput analysis devices. The explosion in data volume, variety, and complexity has required modern supercomputers to improve the sustained performance continuously and bring HPC to the extreme scale. The most significant changes at the extreme scale come from architectural changes in the underlying hardware platforms. On-chip parallelism is dramatically increased to improve performance as power and cooling constraints have limited increases in microprocessor clock speeds. The number of cores is expected to be hundreds or even thousands per chip. Total concurrency must rise by up to five orders of magnitude on the extreme scale systems. Meanwhile, the available memory and memory bandwidth will not scale by the same order of magnitude. The disparity of growth between computing and memory means that memory capacity and bandwidth per core even less than today’s petascale machines. This has led to deeper memory and storage hierarchies to keep most of relevant data for a program close to the processing logic. The phenomenal increase in datasets and the rapid advance of computing architectures on extreme scale high performance computing systems bring critical challenges to data organization and management. The software solution that addresses these challenges, however, has significantly lagged behind. This dissertation research focuses on designing and developing an innovative coordinated data movement methodology that can substantially improve the performance and energy efficiency of high performance computing. This new methodology provides a coordinated architecture that organizes data movement to achieve desired parallelism, locality, and scalability for extreme scale systems. It adds features to the programming environment and develops an execution model that offers effective methods to explicitly arrange data movement. This dissertation advances the understandings of scientific applications’ I/O activities and further unleashes the power of high performance computing.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A New Method for Ranking Extreme Efficient DMUs Based on Changing the Reference Set with Using L2 - Norm

The purpose of this study is to utilize a new method for ranking extreme efficient decision making units (DMUs) based upon the omission of these efficient DMUs from reference set of inefficient and non-extreme efficient DMUs in data envelopment analysis (DEA) models with constant and variable returns to scale. In this method, an L2- norm is used and it is believed that it doesn't have any e...

متن کامل

Performance of communication patterns for extreme-scale analysis and visualization

Efficient data movement is essential for extreme-scale parallel visualization and analysis algorithms. In this research, we benchmark and optimize the performance of collective and point-to-point communication patterns for data-parallel visualization of scalar and vector data. Two such communication patterns are global reduction and local nearest-neighbor communication. We implement scalable al...

متن کامل

Active flash: towards energy-efficient, in-situ data analytics on extreme-scale machines

Modern scientific discovery is increasingly driven by large-scale supercomputing simulations, followed by data analysis tasks. These data analyses are either performed offline, on smaller-scale clusters, or in-situ, on the supercomputer itself. Both of these strategies are rife with storage and I/O bottlenecks, energy inefficiencies due to increased data movement, and increased time to solution...

متن کامل

An approach to rank efficient DMUs in DEA based on combining Manhattan and infinity norms

In many applications, discrimination among decision making units (DMUs) is a problematic technical task procedure to decision makers in data envelopment analysis (DEA). The DEA models unable to discriminate between extremely efficient DMUs. Hence, there is a growing interest in improving discrimination power in DEA yet. The aim of this paper is ranking extreme efficient DMUs in DEA based on exp...

متن کامل

Reducing Data Movement Costs Using Energy-Efficient, Active Computation on SSD

Modern scientific discovery often involves running complex application simulations on supercomputers, followed by a sequence of data analysis tasks on smaller clusters. This offline approach suffers from significant data movement costs such as redundant I/O, storage bandwidth bottleneck, and wasted CPU cycles, all of which contribute to increased energy consumption and delayed end-toend perform...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015